Asymptotic Ancillarity and Conditional Inference for Stochastic Processes
نویسندگان
چکیده
منابع مشابه
Asymptotic Expansions for Stochastic Processes
The central limit theorems are the basis of the large sample statistics. In estimation theory, the asymptotic efficiency is evaluated by the asymptotic variance of estimators, and in testing statistical hypotheses, the critical region of a test is determined by the normal approximation. Though asymptotic properties of statistics are based on central limit theorems, the accuracy of their approxi...
متن کاملStochastic Conditional Intensity Processes
In this article, we introduce the so-called stochastic conditional intensity (SCI) model by extending Russell’s (1999) autoregressive conditional intensity (ACI) model by a latent common dynamic factor that jointly drives the individual intensity components. We show by simulations that the proposed model allows for a wide range of (cross-)autocorrelation structures in multivariate point process...
متن کاملAsymptotic Inference for Partially Observed Branching Processes
We consider the problem of estimation in a partially observed discrete-time Galton– Watson branching process, focusing on the first twomoments of the offspring distribution. Our study is motivated by modelling the counts of new cases at the onset of a stochastic epidemic, allowing for the facts that only a part of the cases is detected, and that the detection mechanism may affect the evolution ...
متن کاملStatistical inference for stochastic processes: concepts and developments in asymptotic theory
1 Frame of the first-order asymptotic decision theory Consider a sequence of statistical experiments ET = (X T ,AT , {P T θ }θ∈Θ) (T ∈ R+). Let θ̂T : X T → Θ be a sequence of estimators of the unknown parameter θ. A basic property θ̂T should have is the consistency : θ̂T →PT θ θ (T → ∞) for every θ ∈ Θ. The analyst should not use any estimator without checking this property. For example, if one us...
متن کاملConditional Risk Minimization for Stochastic Processes
We study the task of learning from non-i.i.d. data. In particular, we aim at learning predictors that minimize the conditional risk for a stochastic process, i.e. the expected loss taking into account the set of training samples observed so far. For non-i.i.d. data, the training set contains information about the upcoming samples, so learning with respect to the conditional distribution can be ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: The Annals of Statistics
سال: 1992
ISSN: 0090-5364
DOI: 10.1214/aos/1176348542